An Empirical Evaluation of HHMM Parsing Time
نویسندگان
چکیده
Current state of the art speech recognition systems use very little structural linguistic information while doing word recognition. Some systems attempt to apply syntactic and semantic analysis to speech, but this is typically done in a pipelined approach, where there is thresholding done in between each stage. It would be advantageous to make use of information about higher level linguistic structure before doing any thresholding, so that uncertainty at different levels (acoustic, word level, syntax, semantics) can all be weighed simultaneously to recover the most likely global outcome. However, the standard CYK parsing algorithm has cubic run-time complexity, which makes it difficult to use in streaming speech applications with unsegmented utterances. Some have proposed frameworks for parsing using time-series models such as hierarchical hidden Markov models (HHMMs), with an architecture similar to that of Murphy and Paskin (2001). These models have been proposed because there are algorithms for recovering a most likely sequence in linear time. However, transforming a grammar into a format usable by time series models like HHMMs increases the constants in the run-time complexity, and the practical effect of these transformations on run-time complexity has not been evaluated thoroughly. This paper describes a system that does parsing using an HHMM, and shows preliminary experimental results on run-time for utterances of vary-
منابع مشابه
Complexity Metrics in an Incremental Right-Corner Parser
Hierarchical HMM (HHMM) parsers make promising cognitive models: while they use a bounded model of working memory and pursue incremental hypotheses in parallel, they still achieve parsing accuracies competitive with chart-based techniques. This paper aims to validate that a right-corner HHMM parser is also able to produce complexity metrics, which quantify a reader’s incremental difficulty in u...
متن کاملHHMM Parsing with Limited Parallelism
Hierarchical Hidden Markov Model (HHMM) parsers have been proposed as psycholinguistic models due to their broad coverage within human-like working memory limits (Schuler et al., 2008) and ability to model human reading time behavior according to various complexity metrics (Wu et al., 2010). But HHMMs have been evaluated previously only with very wide beams of several thousand parallel hypothes...
متن کاملA Syntactic Time-Series Model for Parsing Fluent and Disfluent Speech
This paper describes an incremental approach to parsing transcribed spontaneous speech containing disfluencies with a Hierarchical Hidden Markov Model (HHMM). This model makes use of the right-corner transform, which has been shown to increase non-incremental parsing accuracy on transcribed spontaneous speech (Miller and Schuler, 2008), using trees transformed in this manner to train the HHMM p...
متن کاملبررسی مقایسهای تأثیر برچسبزنی مقولات دستوری بر تجزیه در پردازش خودکار زبان فارسی
In this paper, the role of Part-of-Speech (POS) tagging for parsing in automatic processing of the Persian language is studied. To this end, the impact of the quality of POS tagging as well as the impact of the quantity of information available in the POS tags on parsing are studied. To reach the goals, three parsing scenarios are proposed and compared. In the first scenario, the parser assigns...
متن کاملAn improved joint model: POS tagging and dependency parsing
Dependency parsing is a way of syntactic parsing and a natural language that automatically analyzes the dependency structure of sentences, and the input for each sentence creates a dependency graph. Part-Of-Speech (POS) tagging is a prerequisite for dependency parsing. Generally, dependency parsers do the POS tagging task along with dependency parsing in a pipeline mode. Unfortunately, in pipel...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2008